Jan Peters 0001

307 publications

32 venues

h-Index 52

Google's h-Index 83

Additional Information

Achievements
2023 - Research.com Computer Science in Germany Leader Award
Google Citations
29932
Name Venue Year citations
Variational Hierarchical Mixtures for Probabilistic Learning of Inverse Dynamics. TPAMI 2024 0
Learning-Based Design and Control for Quadrupedal Robots With Parallel-Elastic Actuators. IEEE Robotics and Automation Letters 2023 0
Hierarchical Policy Blending as Inference for Reactive Robot Control. ICRA 2023 0
Safe Reinforcement Learning of Dynamic High-Dimensional Robotic Tasks: Navigation, Manipulation, Interaction. ICRA 2023 0
Start State Selection for Control Policy Learning from Optimal Trajectories. ICRA 2023 0
SE(3)-DiffusionFields: Learning smooth cost functions for joint grasp and motion optimization through diffusion. ICRA 2023 0
Clustering of Motion Trajectories by a Distance Measure Based on Semantic Features. Humanoids 2023 0
Model-Based Uncertainty in Value Functions. AISTATS 2023 0
Disentangling Interaction Using Maximum Entropy Reinforcement Learning in Multi-Agent Systems. ECAI 2023 0
Accelerating Motion Planning via Optimal Transport. NIPS/NeurIPS 2023 0
Diminishing Return of Value Expansion Methods in Model-Based Reinforcement Learning. ICLR 2023 0
LS-IQ: Implicit Reward Regularization for Inverse Reinforcement Learning. ICLR 2023 0
Placing by Touching: An Empirical Study on the Importance of Tactile Sensing for Precise Object Placing. IROS 2023 0
Motion Planning Diffusion: Learning and Planning of Robot Motions with Diffusion Models. IROS 2023 0
Digital Twin of a Driver-in-the-Loop Race Car Simulation With Contextual Reinforcement Learning. IEEE Robotics and Automation Letters 2023 0
PAC-Bayes Bounds for Bandit Problems: A Survey and Experimental Comparison. TPAMI 2023 0
Continuous-Time Fitted Value Iteration for Robust Policies. TPAMI 2023 0
A Deterministic Approximation to Neural SDEs. TPAMI 2023 0
Composable energy policies for reactive motion generation and reinforcement learning. IJRR 2023 0
Combining physics and deep learning to learn continuous-time dynamics models. IJRR 2023 0
Learning to Control Highly Accelerated Ballistic Movements on Muscular Robots. Robotics and Autonomous Systems 2023 0
MILD: Multimodal Interactive Latent Dynamics for Learning Human-Robot Interaction. Humanoids 2022 0
HEBO: An Empirical Study of Assumptions in Bayesian Optimisation. JAIR 2022 34
Active Exploration for Robotic Manipulation. IROS 2022 0
Learning Implicit Priors for Motion Optimization. IROS 2022 7
Curriculum Reinforcement Learning via Constrained Optimal Transport. ICML 2022 5
Dimensionality Reduction and Prioritized Exploration for Policy Search. AISTATS 2022 1
Boosted Curriculum Reinforcement Learning. ICLR 2022 2
Graph-based Reinforcement Learning meets Mixed Integer Programs: An application to 3D robot assembly discovery. IROS 2022 0
Monte-Carlo Robot Path Planning. IEEE Robotics and Automation Letters 2022 1
Controlling the Cascade: Kinematic Planning for N-ball Toss Juggling. IROS 2022 1
Regularized Deep Signed Distance Fields for Reactive Motion Generation. IROS 2022 7
Improving Sample Efficiency of Example-Guided Deep Reinforcement Learning for Bipedal Walking. Humanoids 2022 0
PAC-Bayesian lifelong learning for multi-armed bandits. DMKD 2022 3
Robot Learning of Mobile Manipulation With Reachability Behavior Priors. IEEE Robotics and Automation Letters 2022 4
Adapting Object-Centric Probabilistic Movement Primitives with Residual Reinforcement Learning. Humanoids 2022 0
Integrated Bi-Manual Motion Generation and Control shaped for Probabilistic Movement Primitives. Humanoids 2022 0
Inferring Smooth Control: Monte Carlo Posterior Policy Iteration with Gaussian Processes. CoRL 2022 0
Information-Theoretic Safe Exploration with Gaussian Processes. NIPS/NeurIPS 2022 0
Learning Stable Vector Fields on Lie Groups. IEEE Robotics and Automation Letters 2022 0
Benchmarking Structured Policies and Policy Optimization for Real-World Dexterous Object Manipulation. IEEE Robotics and Automation Letters 2022 0
Continuous Action Reinforcement Learning From a Mixture of Interpretable Experts. TPAMI 2022 0
Batch Reinforcement Learning With a Nonparametric Off-Policy Policy Gradient. TPAMI 2022 0
Composable Energy Policies for Reactive Motion Generation and Reinforcement Learning. RSS 2021 14
Value Iteration in Continuous Actions, States and Time. ICML 2021 8
Directed Acyclic Graph Neural Network for Human Motion Prediction. ICRA 2021 3
Gaussian Approximation for Bias Reduction in Q-Learning. JMLR 2021 4
SKID RAW: Skill Discovery From Raw Trajectories. IEEE Robotics and Automation Letters 2021 11
Learn2Assemble with Structured Representations and Search for Robotic Architectural Construction. CoRL 2021 14
Neural Posterior Domain Randomization. CoRL 2021 11
A Probabilistic Interpretation of Self-Paced Learning with Applications to Reinforcement Learning. JMLR 2021 8
Latent Derivative Bayesian Last Layer Networks. AISTATS 2021 5
An Empirical Analysis of Measure-Valued Derivatives for Policy Gradients. IJCNN 2021 5
Efficient and Reactive Planning for High Speed Robot Air Hockey. IROS 2021 6
Robot Reinforcement Learning on the Constraint Manifold. CoRL 2021 16
Model Predictive Actor-Critic: Accelerating Robot Skill Acquisition with Deep Reinforcement Learning. ICRA 2021 22
Convex optimization with an interpolation-based projection and its application to deep learning. MLJ 2021 0
Robust Value Iteration for Continuous Control Tasks. RSS 2021 11
Learning Human-like Hand Reaching for Human-Robot Handshaking. ICRA 2021 5
Convex Regularization in Monte-Carlo Tree Search. ICML 2021 0
A Variational Infinite Mixture for Probabilistic Inverse Dynamics Learning. ICRA 2021 0
Contextual Latent-Movements Off-Policy Optimization for Robotic Manipulation Skills. ICRA 2021 0
Differentiable Physics Models for Real-world Offline Model-based Reinforcement Learning. ICRA 2021 0
Data-Efficient Domain Randomization With Bayesian Optimization. IEEE Robotics and Automation Letters 2021 0
Assessing Transferability From Simulation to Reality for Reinforcement Learning. TPAMI 2021 0
MushroomRL: Simplifying Reinforcement Learning Research. JMLR 2021 0
A Probabilistic Framework for Imitating Human Race Driver Behavior. IEEE Robotics and Automation Letters 2020 6
Learning attribute grammars for movement primitive sequencing. IJRR 2020 7
Model-Based Quality-Diversity Search for Efficient Robot Learning. IROS 2020 15
Multi-Sensor Next-Best-View Planning as Matroid-Constrained Submodular Maximization. IEEE Robotics and Automation Letters 2020 8
ImitationFlow: Learning Deep Stable Stochastic Dynamic Systems by Normalizing Flows. IROS 2020 24
Self-Paced Deep Reinforcement Learning. NIPS/NeurIPS 2020 22
Deep Adversarial Reinforcement Learning for Object Disentangling. IROS 2020 1
Probabilistic Approach to Physical Object Disentangling. IEEE Robotics and Automation Letters 2020 3
Multi-agent active information gathering in discrete and continuous-state decentralized POMDPs by policy graph improvement. Autonomous Agents and Multi-Agent Systems 2020 9
High Acceleration Reinforcement Learning for Real-World Juggling with Binary Rewards. CoRL 2020 12
Redundancy resolution under hard joint constraints: a generalized approach to rank updates. IROS 2020 1
Learning Hierarchical Acquisition Functions for Bayesian Optimization. IROS 2020 0
Learning Control Policies from Optimal Trajectories. ICRA 2020 1
Numerical Quadrature for Probabilistic Policy Search. TPAMI 2020 9
Underactuated Waypoint Trajectory Optimization for Light Painting Photography. ICRA 2020 1
Plucking Motions for Tea Harvesting Robots Using Probabilistic Movement Primitives. IEEE Robotics and Automation Letters 2020 12
A Nonparametric Off-Policy Policy Gradient. AISTATS 2020 11
Sharing Knowledge in Multi-Task Deep Reinforcement Learning. ICLR 2020 66
Bayesian Online Prediction of Change Points. UAI 2020 0
Generalized Mean Estimation in Monte-Carlo Tree Search. IJCAI 2020 0
Real Time Trajectory Prediction Using Deep Conditional Generative Models. IEEE Robotics and Automation Letters 2020 0
Multimodal Uncertainty Reduction for Intention Recognition in Human-Robot Interaction. IROS 2019 18
Reinforcement Learning of Trajectory Distributions: Applications in Assisted Teleoperation and Motion Planning. IROS 2019 7
TD-regularized actor-critic methods. MLJ 2019 0
Experience Reuse with Probabilistic Movement Primitives. IROS 2019 9
Deep Lagrangian Networks for end-to-end learning of energy-based control for under-actuated systems. IROS 2019 42
Switching Linear Dynamics for Variational Bayes Filtering. ICML 2019 37
Building a Library of Tactile Skills Based on FingerVision. Humanoids 2019 6
Generalized Multiple Correlation Coefficient as a Similarity Measurement between Trajectories. IROS 2019 4
Self-Paced Contextual Reinforcement Learning. CoRL 2019 30
Reinforcement learning of motor skills using Policy Search and human corrective advice. IJRR 2019 10
Chance-Constrained Trajectory Optimization for Non-linear Systems with Unknown Stochastic Dynamics. IROS 2019 5
Receding Horizon Curiosity. CoRL 2019 8
Compatible natural gradient policy search. MLJ 2019 0
Learning Replanning Policies With Direct Policy Search. IEEE Robotics and Automation Letters 2019 4
Projections for Approximate Policy Iteration Algorithms. ICML 2019 8
Learning Intention Aware Online Adaptation of Movement Primitives. IEEE Robotics and Automation Letters 2019 12
Information Gathering in Decentralized POMDPs by Policy Graph Improvement. AAMAS 2019 6
Entropic Risk Measure in Policy Search. IROS 2019 18
Stochastic Optimal Control as Approximate Input Inference. CoRL 2019 16
HJB Optimal Feedback Control with Deep Differential Value Functions and Action Constraints. CoRL 2019 12
Exploration Driven by an Optimistic Bellman Equation. IJCNN 2019 1
Local Online Motor Babbling: Learning Motor Abundance of a Musculoskeletal Robot Arm IROS 2019 0
Learning to Serve: An Experimental Study for a New Learning From Demonstrations Framework. IEEE Robotics and Automation Letters 2019 0
Online Learning of an Open-Ended Skill Library for Collaborative Tasks. Humanoids 2018 16
Inverse Reinforcement Learning via Nonparametric Spatio-Temporal Subgoal Modeling. JMLR 2018 7
Inducing Probabilistic Context-Free Grammars for the Sequencing of Movement Primitives. ICRA 2018 4
Using probabilistic movement primitives in robotics. Autonomous Robots 2018 138
Control of Musculoskeletal Systems Using Learned Dynamics Models. IEEE Robotics and Automation Letters 2018 13
Mixture of Attractors: A Novel Movement Primitive Representation for Learning Motor Skills From Demonstrations. IEEE Robotics and Automation Letters 2018 14
Domain Randomization for Simulation-Based Policy Optimization with Transferability Assessment. CoRL 2018 38
Regularizing Reinforcement Learning with State Abstraction. IROS 2018 11
Sample and Feedback Efficient Hierarchical Reinforcement Learning from Human Preferences. ICRA 2018 19
Learning Coupled Forward-Inverse Models with Combined Prediction Errors. ICRA 2018 2
Approximate Value Iteration Based on Numerical Quadrature. IEEE Robotics and Automation Letters 2018 2
Online optimal trajectory generation for robot table tennis. Robotics and Autonomous Systems 2018 45
Utilizing Human Feedback in POMDP Execution and Specification. Humanoids 2018 6
A kernel-based approach to learning contact distributions for robot manipulation tasks. Autonomous Robots 2018 12
PIPPS: Flexible Model-Based Policy Search Robust to the Curse of Chaos. ICML 2018 0
Model-Free Trajectory-based Policy Optimization with Monotonic Improvement. JMLR 2018 0
Anticipatory action selection for human-robot table tennis. Artificial Intelligence 2017 31
Learning inverse dynamics models in O(n) time with LSTM networks. Humanoids 2017 47
Whole-body multi-contact motion in humans and humanoids: Advances of the CoDyCo European project. Robotics and Autonomous Systems 2017 19
Layered direct policy search for learning hierarchical skills. ICRA 2017 17
Stability of Controllers for Gaussian Process Dynamics. JMLR 2017 28
Probabilistic Prioritization of Movement Primitives. IEEE Robotics and Automation Letters 2017 21
Efficient online adaptation with stochastic recurrent neural networks. Humanoids 2017 3
A comparison of distance measures for learning nonparametric motor skill libraries. Humanoids 2017 3
Online Learning with Stochastic Recurrent Neural Networks using Intrinsic Motivation Signals. CoRL 2017 4
Hybrid control trajectory optimization under uncertainty. IROS 2017 13
Context-driven movement primitive adaptation. ICRA 2017 5
Non-parametric Policy Search with Limited Information Loss. JMLR 2017 26
Learning movement primitive libraries through probabilistic segmentation. IJRR 2017 49
Model-based contextual policy search for data-efficient generalization of robot skills. Artificial Intelligence 2017 72
A learning-based shared control architecture for interactive task execution. ICRA 2017 40
Local Bayesian Optimization of Motor Skills. ICML 2017 16
Generalized exploration in policy search. MLJ 2017 0
Empowered skills. ICRA 2017 4
A Comparison of Autoregressive Hidden Markov Models for Multimodal Manipulations With Variable Masses. IEEE Robotics and Automation Letters 2017 1
Guiding Trajectory Optimization by Demonstrated Distributions. IEEE Robotics and Automation Letters 2017 39
Goal-driven dimensionality reduction for reinforcement learning. IROS 2017 12
State-Regularized Policy Search for Linearized Dynamical Systems. ICAPS 2017 6
Active Incremental Learning of Robot Movement Primitives. CoRL 2017 42
Policy Search with High-Dimensional Context Variables. AAAI 2017 0
Phase estimation for fast action recognition and trajectory generation in human-robot collaboration. IJRR 2017 0
Probabilistic movement primitives for coordination of multiple human-robot collaborative tasks. Autonomous Robots 2017 0
Active tactile object exploration with Gaussian processes. IROS 2016 66
Demonstration based trajectory optimization for generalizable robot motions. Humanoids 2016 38
Deep spiking networks for model-based planning in humanoids. Humanoids 2016 10
Incremental imitation learning of context-dependent motor skills. Humanoids 2016 17
Probabilistic inference for determining options in reinforcement learning. MLJ 2016 0
Stable reinforcement learning with autoencoders for tactile and visual data. IROS 2016 135
Movement primitives with multiple phase parameters. ICRA 2016 6
Stability of Controllers for Gaussian Process Forward Models. ICML 2016 34
Jointly learning trajectory generation and hitting point prediction in robot table tennis. Humanoids 2016 29
Using probabilistic movement primitives for striking movements. Humanoids 2016 19
Learning soft task priorities for control of redundant robots. ICRA 2016 35
A new trajectory generation framework in robotic table tennis. IROS 2016 5
Probabilistic decomposition of sequential force interaction tasks into Movement Primitives. IROS 2016 10
Hierarchical Relative Entropy Policy Search. JMLR 2016 0
Acquiring and Generalizing the Embodiment Mapping From Human Observations to Robot Skills. IEEE Robotics and Automation Letters 2016 21
A lightweight robotic arm with pneumatic muscles for robot learning. ICRA 2016 32
Catching heuristics are optimal control policies. NIPS/NeurIPS 2016 29
Manifold Gaussian Processes for regression. IJCNN 2016 0
Probabilistic progress prediction and sequencing of concurrent movement primitives. IROS 2015 9
Learning optimal striking points for a ping-pong playing robot. IROS 2015 22
Learning multiple collaborative tasks with a mixture of Interaction Primitives. ICRA 2015 99
Learning inverse dynamics models with contacts. ICRA 2015 56
Learning of Non-Parametric Control Policies with High-Dimensional State Features. AISTATS 2015 34
Evaluation of tactile feature extraction for interactive object recognition. Humanoids 2015 58
Reinforcement learning vs human programming in tetherball robot games. IROS 2015 15
Learning movement primitive attractor goals and sequential skills from kinesthetic demonstrations. Robotics and Autonomous Systems 2015 55
Optimizing robot striking movement primitives with Iterative Learning Control. Humanoids 2015 4
Active reward learning with a novel acquisition function. Autonomous Robots 2015 51
Towards learning hierarchical skills for multi-phase manipulation tasks. ICRA 2015 109
Learning robot in-hand manipulation with tactile features. Humanoids 2015 137
Stabilizing novel objects by learning to predict tactile slip. IROS 2015 86
A Probabilistic Framework for Semi-autonomous Robots Based on Interaction Primitives with Phase Estimation. ISRR 2015 15
Extracting low-dimensional control variables for movement primitives. ICRA 2015 44
First-person tele-operation of a humanoid robot. Humanoids 2015 27
Combined pose-wrench and state machine representation for modeling Robotic Assembly Skills. IROS 2015 15
A comparison of contact distribution representations for learning to predict object interactions. Humanoids 2015 5
Learning torque control in presence of contacts using tactile sensing from robot skin. Humanoids 2015 13
Model-free Probabilistic Movement Primitives for physical interaction. IROS 2015 17
Probabilistic segmentation applied to an assembly task. Humanoids 2015 38
Learning motor skills from partially observed movements executed at different speeds. IROS 2015 25
Policy Evaluation with Temporal Differences: A Survey and Comparison (Extended Abstract). ICAPS 2015 0
Model-Based Relative Entropy Stochastic Search. NIPS/NeurIPS 2015 0
Policy search for learning robot control using sparse data. ICRA 2014 18
Learning to sequence movement primitives from demonstrations. IROS 2014 45
Tools for simulating humanoid robot dynamics: A survey based on user feedback. Humanoids 2014 64
Learning robot tactile sensing for object manipulation. IROS 2014 67
Special issue on autonomous grasping and manipulation. Autonomous Robots 2014 12
Interaction primitives for human-robot cooperation tasks. ICRA 2014 199
Learning interaction for collaborative tasks with probabilistic movement primitives. Humanoids 2014 88
An experimental comparison of Bayesian optimization for bipedal locomotion. ICRA 2014 72
Latent space policy search for robotics. IROS 2014 25
Dimensionality reduction for probabilistic movement primitives. Humanoids 2014 33
Predicting object interactions from contact distributions. IROS 2014 20
Multi-task policy search for robotics. ICRA 2014 111
Robust policy updates for stochastic optimal control. Humanoids 2014 8
Active Reward Learning. RSS 2014 98
Natural evolution strategies. JMLR 2014 9
Policy Search for Path Integral Control. ECML/PKDD 2014 41
Generalizing pouring actions between objects using warped parameters. Humanoids 2014 33
Learning to predict phases of manipulation tasks as hidden states. ICRA 2014 45
Sample-based informationl-theoretic stochastic optimal control. ICRA 2014 23
Policy evaluation with temporal differences: a survey and comparison. JMLR 2014 0
Towards Robot Skill Learning: From Simple Skills to Table Tennis. ECML/PKDD 2013 39
Probabilistic interactive segmentation for anthropomorphic robots in cluttered environments. Humanoids 2013 10
Probabilistic Movement Primitives. NIPS/NeurIPS 2013 444
Reinforcement learning in robotics: A survey. IJRR 2013 2459
Learning sequential motor tasks. ICRA 2013 49
A probabilistic approach to robot trajectory generation. Humanoids 2013 20
Feedback error learning for rhythmic motor primitives. ICRA 2013 7
Autonomous reinforcement learning with hierarchical REPS. IJCNN 2013 5
Learning responsive robot behavior by imitation. IROS 2013 39
Model-based imitation learning by probabilistic trajectory matching. ICRA 2013 46
Data-Efficient Generalization of Robot Skills with Contextual Policy Search. AAAI 2013 124
Alignment-based transfer learning for robot models. IJCNN 2013 65
Probabilistic movement modeling for intention inference in human-robot interaction. IJRR 2013 0
Learning to select and generalize striking movements in robot table tennis. IJRR 2013 0
A kernel-based approach to direct action perception. ICRA 2012 85
Algorithms for Learning Markov Field Policies. NIPS/NeurIPS 2012 0
Generalization of human grasping for multi-fingered robot hands. IROS 2012 98
Learning concurrent motor skills in versatile solution spaces. IROS 2012 45
Learning throwing and catching skills. IROS 2012 10
Maximally informative interaction learning for scene exploration. IROS 2012 43
Probabilistic Modeling of Human Movements for Intention Inference. RSS 2012 52
Robot Skill Learning. ECAI 2012 9
Reinforcement learning to adjust parametrized motor primitives to new situations. Autonomous Robots 2012 0
Structured Apprenticeship Learning. ECML/PKDD 2012 26
Learning tracking control with forward models. ICRA 2012 10
Toward fast policy search for learning legged locomotion. IROS 2012 33
Point cloud completion using extrusions. Humanoids 2012 28
A brain-robot interface for studying motor learning after stroke. IROS 2012 20
Hierarchical Relative Entropy Policy Search. AISTATS 2012 0
Learning visual representations for perception-action systems. IJRR 2011 41
Trajectory planning for optimal robot catching in real-time. ICRA 2011 112
Modeling Opponent Actions for Table-Tennis Playing Robot. AAAI 2011 1
Model learning for robot control: a survey. Cognitive Processing 2011 455
A flexible hybrid framework for modeling complex manipulation tasks. ICRA 2011 8
Learning elementary movements jointly with a higher level task. IROS 2011 16
A Non-Parametric Approach to Dynamic Programming. NIPS/NeurIPS 2011 19
Learning robot grasping from 3-D images with Markov Random Fields. IROS 2011 45
Relative Entropy Inverse Reinforcement Learning. AISTATS 2011 315
Balancing Safety and Exploitability in Opponent Modeling. AAAI 2011 19
Learning inverse kinematics with structured prediction. IROS 2011 49
Learning task-space tracking control with kernels. IROS 2011 6
Learning anticipation policies for robot table tennis. IROS 2011 19
Reinforcement Learning to Adjust Robot Movements to New Situations. IJCAI 2011 0
Policy search for motor primitives in robotics. MLJ 2011 0
Movement extraction by detecting dynamics switches and repetitions. NIPS/NeurIPS 2010 60
Learning table tennis with a Mixture of Motor Primitives. Humanoids 2010 97
Combining active learning and reactive control for robot grasping. Robotics and Autonomous Systems 2010 164
Using model knowledge for learning inverse dynamics. ICRA 2010 171
Reinforcement Learning to adjust Robot Movements to New Situations. RSS 2010 203
A biomimetic approach to robot table tennis. IROS 2010 113
Learning probabilistic discriminative models of grasp affordances under limited supervision. IROS 2010 21
Relative Entropy Policy Search. AAAI 2010 605
Switched Latent Force Models for Movement Segmentation. NIPS/NeurIPS 2010 37
Incremental Sparsification for Real-time Online Model Learning. AISTATS 2010 25
Movement templates for learning of hitting and batting. ICRA 2010 169
Learning Visual Representations for Interactive Systems. ISRR 2009 11
An Expectation Maximization Algorithm for Continuous Markov Decision Processes with Arbitrary Reward. AISTATS 2009 33
Learning complex motions by sequencing simpler motion templates. ICML 2009 47
Efficient Sample Reuse in EM-Based Policy Search. ECML/PKDD 2009 17
Active learning using mean shift optimization for robot grasping. IROS 2009 35
Sparse online model learning for robot control with support vector regression. IROS 2009 31
Guest editorial: Special issue on robot learning, Part A. Autonomous Robots 2009 2
Learning motor primitives for robotics. ICRA 2009 236
Towards Motor Skill Learning for Robotics. ISRR 2009 0
Guest editorial: Special issue on robot learning, Part B. Autonomous Robots 2009 0
Real-time learning of resolved velocity control on a Mitsubishi PA-10. ICRA 2008 7
Local Gaussian Process Regression for Real Time Online Model Learning. NIPS/NeurIPS 2008 242
Local Gaussian process regression for real-time model-based robot control. IROS 2008 138
Learning to Control in Operational Space. IJRR 2008 183
Operational Space Control: A Theoretical and Empirical Comparison. IJRR 2008 353
Policy Gradients with Parameter-Based Exploration for Control. ICANN 2008 70
Fitted Q-iteration by Advantage Weighted Regression. NIPS/NeurIPS 2008 58
Model-Based Reinforcement Learning with Continuous States and Actions. ESANN 2008 23
Policy Search for Motor Primitives in Robotics. NIPS/NeurIPS 2008 712
Learning perceptual coupling for motor primitives. IROS 2008 90
Episodic Reinforcement Learning by Logistic Reward-Weighted Regression. ICANN 2008 10
Using Bayesian Dynamical Systems for Motion Template Libraries. NIPS/NeurIPS 2008 26
Learning Inverse Dynamics: a Comparison. ESANN 2008 89
Adaptive Importance Sampling with Automatic Model Selection in Value Function Approximation. AAAI 2008 0
A unifying framework for robot control with redundant DOFs. Autonomous Robots 2008 0
Reinforcement Learning for Operational Space Control. ICRA 2007 28
Applying the Episodic Natural Actor-Critic Architecture to Motor Primitive Learning. ESANN 2007 28
Towards compliant humanoids-an experimental assessment of suitable task space position/orientation controllers. IROS 2007 5
Reinforcement learning by reward-weighted regression for operational space control. ICML 2007 197
Solving Deep Memory POMDPs with Recurrent Policy Gradients. ICANN 2007 159
Policy Gradient Methods for Robotics. IROS 2006 559
Reinforcement Learning for Parameterized Motor Primitives. IJCNN 2006 30
Learning Operational Space Control. RSS 2006 22
A Bayesian Approach to Nonlinear Parameter Identification for Rigid Body Dynamics. RSS 2006 63
A unifying methodology for the control of robotic systems. IROS 2005 42
Comparative experiments on task space control with redundancy resolution. IROS 2005 82
Natural Actor-Critic. ECML/PKDD 2005 0
Learning Movement Primitives. ISRR 2003 0
Copyright ©2019 Universität Würzburg

Impressum | Privacy | FAQ